sparse reward and multiple goal
PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals
Learning with sparse rewards remains a significant challenge in reinforcement learning (RL), especially when the aim is to train a policy capable of achieving multiple different goals. To date, the most successful approaches for dealing with multi-goal, sparse reward environments have been model-free RL algorithms. In this work we propose PlanGAN, a model-based algorithm specifically designed for solving multi-goal tasks in environments with sparse rewards. Our method builds on the fact that any trajectory of experience collected by an agent contains useful information about how to achieve the goals observed during that trajectory. We use this to train an ensemble of conditional generative models (GANs) to generate plausible trajectories that lead the agent from its current state towards a specified goal. We then combine these imagined trajectories into a novel planning algorithm in order to achieve the desired goal as efficiently as possible. The performance of PlanGAN has been tested on a number of robotic navigation/manipulation tasks in comparison with a range of model-free reinforcement learning baselines, including Hindsight Experience Replay. Our studies indicate that PlanGAN can achieve comparable performance whilst being around 4-8 times more sample efficient.
Review for NeurIPS paper: PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals
This paper proposes using an ensemble of GANs to learn a goal-conditioned forward model of trajectories for use in planning. The model is trained using a variant of hindsight experience replay, resulting in an agent that can succeed at sparse goal-conditioned tasks with much better data efficiency than model-free approaches. All reviewers highlighted the impressiveness of the experimental results, with R1 and R2 finding the approach very interesting, and R3 and R4 indicating the potential impact and interest this work will have. I agree that this paper will likely be of broad interest to the RL community at NeurIPS and therefore recommend acceptance. However, several reviewers also noted the lack of comparison to other model-based approaches.
PlanGAN: Model-based Planning With Sparse Rewards and Multiple Goals
Learning with sparse rewards remains a significant challenge in reinforcement learning (RL), especially when the aim is to train a policy capable of achieving multiple different goals. To date, the most successful approaches for dealing with multi-goal, sparse reward environments have been model-free RL algorithms. In this work we propose PlanGAN, a model-based algorithm specifically designed for solving multi-goal tasks in environments with sparse rewards. Our method builds on the fact that any trajectory of experience collected by an agent contains useful information about how to achieve the goals observed during that trajectory. We use this to train an ensemble of conditional generative models (GANs) to generate plausible trajectories that lead the agent from its current state towards a specified goal.
Shadow Robot: AI Algorithms Bring Robot Hands One Step Closer to Human
The Shadow Robot Dexterous Hand is a robot hand, with size, shape, and movement capabilities similar to those of a human hand. To give the robotic hand the ability to learn how to manipulate objects researchers from WMG, University of Warwick, have developed new AI algorithms. Robot hands can be used in many applications, such as manufacturing, surgery and dangerous activities like nuclear decommissioning. For instance, robotic hands can be very useful in computer assembly where assembling microchips requires a level of precision that only human hands can currently achieve. Thanks to the utilization of robot hands in assembly lines, higher productivity may be achieved whilst securing reduced exposure from work risk situations to human workers.